Recent studies have found that pain in infancy has a significant impact on infant development, including psychological problems, possible brain injury, and pain sensitivity in adulthood. However, due to the lack of specialists and the fact that infants are unable to express verbally their experience of pain, it is difficult to assess infant pain. Most existing infant pain assessment systems directly apply adult methods to infants ignoring the differences between infant expressions and adult expressions. Meanwhile, as the study of facial action coding system continues to advance, the use of action units (AUs) opens up new possibilities for expression recognition and pain assessment. In this paper, a novel AuE-IPA method is proposed for assessing infant pain by leveraging different engagement levels of AUs. First, different engagement levels of AUs in infant pain are revealed, by analyzing the class activation map of an end-to-end pain assessment model. The intensities of top-engaged AUs are then used in a regression model for achieving automatic infant pain assessment. The model proposed is trained and experimented on YouTube Immunization dataset, YouTube Blood Test dataset, and iCOPEVid dataset. The experimental results show that our AuE-IPA method is more applicable to infants and possesses stronger generalization ability than end-to-end assessment model and the classic PSPI metric.
translated by 谷歌翻译
Machine learning has emerged recently as a powerful tool for predicting properties of quantum many-body systems. For many ground states of gapped Hamiltonians, generative models can learn from measurements of a single quantum state to reconstruct the state accurately enough to predict local observables. Alternatively, kernel methods can predict local observables by learning from measurements on different but related states. In this work, we combine the benefits of both approaches and propose the use of conditional generative models to simultaneously represent a family of states, by learning shared structures of different quantum states from measurements. The trained model allows us to predict arbitrary local properties of ground states, even for states not present in the training data, and without necessitating further training for new observables. We numerically validate our approach (with simulations of up to 45 qubits) for two quantum many-body problems, 2D random Heisenberg models and Rydberg atom systems.
translated by 谷歌翻译
学习在线推荐模型的关键挑战之一是时间域移动,这会导致培训与测试数据分布之间的不匹配以及域的概括错误。为了克服,我们建议学习一个未来的梯度生成器,该生成器可以预测培训未来数据分配的梯度信息,以便可以对建议模型进行培训,就像我们能够展望其部署的未来一样。与批处理更新相比,我们的理论表明,所提出的算法达到了较小的时间域概括误差,该误差通过梯度变异项在局部遗憾中衡量。我们通过与各种代表性基线进行比较来证明经验优势。
translated by 谷歌翻译
在分布式深度学习的背景下,陈旧的权重或梯度的问题可能导致算法性能差。这个问题通常通过延迟耐受算法来解决,并在目标函数和步进尺寸上有一些温和的假设。在本文中,我们提出了一种不同的方法来开发一种新算法,称为$ \ textbf {p} $ redicting $ \ textbf {c} $ lipping $ \ textbf {a} $ synchronous $ \ textbf {s} textbf {g} $ radient $ \ textbf {d} $ escent(aka,pc-asgd)。具体而言,PC -ASGD有两个步骤 - $ \ textIt {预测步骤} $利用泰勒扩展利用梯度预测来减少过时的权重的稳固性,而$ \ textit {clivipping step} $选择性地降低了过时的权重,以减轻过时的权重他们的负面影响。引入权衡参数以平衡这两个步骤之间的影响。从理论上讲,考虑到平滑的物镜函数弱键和非凸的延迟延迟的延迟,我们介绍了收敛速率。还提出了一种实用的PC-ASGD变体,即采用条件来帮助确定权衡参数。对于经验验证,我们在两个基准数据集上使用两个深神经网络体系结构演示了该算法的性能。
translated by 谷歌翻译
体育游戏摘要旨在根据实时评论生成体育新闻。该任务吸引了广泛的研究关注,但由于缺乏相应的英语数据集,但仍未探索。因此,在本文中,我们发布了第一个英语体育游戏摘要数据集的目标。具体而言,目标有103个评论新对,评论和新闻的平均长度分别为2724.9和476.3个字。此外,为了支持半监督环境中的研究,目标还提供了2,160个未标记的评论文件。基于我们的目标,我们建立和评估了几个基线,包括提取性和抽象基线。实验结果表明,此任务的挑战仍然存在。我们希望我们的工作能够促进体育游戏总结的研究。该数据集已在https://github.com/krystalan/goal上发布。
translated by 谷歌翻译
无监督域适应(UDA)的绝大多数现有算法都集中在以一次性的方式直接从标记的源域调整到未标记的目标域。另一方面,逐渐的域适应性(GDA)假设桥接源和目标的$(t-1)$未标记的中间域,并旨在通过利用中间的路径在目标域中提供更好的概括。在某些假设下,Kumar等人。 (2020)提出了一种简单的算法,逐渐自我训练,以及按$ e^{o(t)} \ left的顺序结合的概括(\ varepsilon_0+o \ of \ left(\ sqrt {log(log(log(t)/n log(t)/n) } \ right)\ right)$对于目标域错误,其中$ \ varepsilon_0 $是源域错误,$ n $是每个域的数据大小。由于指数因素,当$ t $仅适中时,该上限变得空虚。在这项工作中,我们在更一般和放松的假设下分析了逐步的自我训练,并证明概括为$ \ varepsilon_0 + o \ left(t \ delta + t/\ sqrt {n} {n} \ right) + \ widetilde { o} \ left(1/\ sqrt {nt} \ right)$,其中$ \ delta $是连续域之间的平均分配距离。与对$ t $作为乘法因素的指数依赖性的现有界限相比,我们的界限仅取决于$ t $线性和添加性。也许更有趣的是,我们的结果意味着存在最佳的$ t $的最佳选择,从而最大程度地减少了概括性错误,并且自然也暗示了一种构造中间域路径的最佳方法,以最大程度地减少累积路径长度$ t \ delta源和目标之间的$。为了证实我们理论的含义,我们检查了对多个半合成和真实数据集的逐步自我训练,这证实了我们的发现。我们相信我们的见解为未来GDA算法设计的途径提供了前进的途径。
translated by 谷歌翻译
域的概括要求在一组训练环境中训练的模型在看不见的测试环境中表现良好。最近,已经提出了一系列算法(例如不变风险最小化(IRM))用于域泛化。但是,Rosenfeld等。 (2021)表明,在简单的线性数据模型中,即使忽略了非凸性问题,IRM及其扩展也无法推广到少于$ d_s+1 $ 1 $训练环境的看不见的环境,其中$ d_s $是$ d_s $的维度。虚假的子空间。在本文中,我们建议通过不变的亚空间恢复(ISR)实现域的概括。我们的第一种算法ISR均值,可以识别出从类条件分布的一阶矩中不变特征所跨越的子空间,并在Rosenfeld等人的数据模型下使用$ D_S+1 $训练环境实现可证明的域概括。 (2021)。我们的第二个算法ISR-COV,使用二阶矩的信息进一步将所需的培训环境数量减少到$ O(1)$。值得注意的是,与IRM不同,我们的算法绕过了非凸度问题并享受全球融合保证。从经验上讲,与IRM相比,我们的ISR可以在合成基准上获得卓越的性能。此外,在三个现实世界的图像和文本数据集上,我们表明这两个ISR可以用作简单但有效的后处理方法,以提高(预)训练有素的模型的最差案例准确性,以抵抗虚假的相关性和群体变化。
translated by 谷歌翻译
目的:提出一种新的基于深度学习的方法,称为RG-NET(重建和生成网络),用于通过向下采样k空间高度加速的MR参数映射,并同时减少所获取的对比度。方法:所提出的框架包括重建模块和生成模块。在先前的帮助下,重建模块从所获取的少数下采样的k空间数据重建MR图像。然后,生成模块从重建的图像中综合剩余的多对比度图像,其中通过对完全采样标签的监督隐式模型被隐式地结合到图像生成中。在不同的加速率下对膝关节和大脑的映射数据进行评估RG-Net。 Cartilage和大脑的区域T1 \ R {HO}进行了分析,以获得RG-Net的性能。结果:RG-Net以高速加速度为17的高质量T1 \ R {Ho}地图。与仅借出k空间的竞争方法相比,我们的框架在T1 \ R {Ho}值中实现了更好的性能分析。我们的方法还提高了胶质瘤患者T1 \ R {Ho}的质量。结论:提出的RG-NET通过欠采样k空间采用新策略并同时减少快速先生参数映射的对比度,可以实现高加速率,同时保持良好的重建质量。我们的框架的生成模块也可以用作其他快速MR参数映射方法的插入模块。关键词:深度学习,卷积神经网络,快速先生参数映射
translated by 谷歌翻译
In this paper, we focus on the problem of feature learning in the presence of scale imbalance for 6-DoF grasp detection and propose a novel approach to especially address the difficulty in dealing with small-scale samples. A Multi-scale Cylinder Grouping (MsCG) module is presented to enhance local geometry representation by combining multi-scale cylinder features and global context. Moreover, a Scale Balanced Learning (SBL) loss and an Object Balanced Sampling (OBS) strategy are designed, where SBL enlarges the gradients of the samples whose scales are in low frequency by apriori weights while OBS captures more points on small-scale objects with the help of an auxiliary segmentation network. They alleviate the influence of the uneven distribution of grasp scales in training and inference respectively. In addition, Noisy-clean Mix (NcM) data augmentation is introduced to facilitate training, aiming to bridge the domain gap between synthetic and raw scenes in an efficient way by generating more data which mix them into single ones at instance-level. Extensive experiments are conducted on the GraspNet-1Billion benchmark and competitive results are reached with significant gains on small-scale cases. Besides, the performance of real-world grasping highlights its generalization ability. Our code is available at https://github.com/mahaoxiang822/Scale-Balanced-Grasp.
translated by 谷歌翻译
In this paper, we propose an end-to-end framework that jointly learns keypoint detection, descriptor representation and cross-frame matching for the task of image-based 3D localization. Prior art has tackled each of these components individually, purportedly aiming to alleviate difficulties in effectively train a holistic network. We design a self-supervised image warping correspondence loss for both feature detection and matching, a weakly-supervised epipolar constraints loss on relative camera pose learning, and a directional matching scheme that detects key-point features in a source image and performs coarse-to-fine correspondence search on the target image. We leverage this framework to enforce cycle consistency in our matching module. In addition, we propose a new loss to robustly handle both definite inlier/outlier matches and less-certain matches. The integration of these learning mechanisms enables end-to-end training of a single network performing all three localization components. Bench-marking our approach on public data-sets, exemplifies how such an end-to-end framework is able to yield more accurate localization that out-performs both traditional methods as well as state-of-the-art weakly supervised methods.
translated by 谷歌翻译